77 research outputs found
Kolmogorov's Structure Functions and Model Selection
In 1974 Kolmogorov proposed a non-probabilistic approach to statistics and
model selection. Let data be finite binary strings and models be finite sets of
binary strings. Consider model classes consisting of models of given maximal
(Kolmogorov) complexity. The ``structure function'' of the given data expresses
the relation between the complexity level constraint on a model class and the
least log-cardinality of a model in the class containing the data. We show that
the structure function determines all stochastic properties of the data: for
every constrained model class it determines the individual best-fitting model
in the class irrespective of whether the ``true'' model is in the model class
considered or not. In this setting, this happens {\em with certainty}, rather
than with high probability as is in the classical case. We precisely quantify
the goodness-of-fit of an individual model with respect to individual data. We
show that--within the obvious constraints--every graph is realized by the
structure function of some data. We determine the (un)computability properties
of the various functions contemplated and of the ``algorithmic minimal
sufficient statistic.''Comment: 25 pages LaTeX, 5 figures. In part in Proc 47th IEEE FOCS; this final
version (more explanations, cosmetic modifications) to appear in IEEE Trans
Inform T
Limit complexities revisited [once more]
The main goal of this article is to put some known results in a common
perspective and to simplify their proofs.
We start with a simple proof of a result of Vereshchagin saying that
equals . Then we use the same argument to prove
similar results for prefix complexity, a priori probability on binary tree, to
prove Conidis' theorem about limits of effectively open sets, and also to
improve the results of Muchnik about limit frequencies. As a by-product, we get
a criterion of 2-randomness proved by Miller: a sequence is 2-random if and
only if there exists such that any prefix of is a prefix of some
string such that . (In the 1960ies this property was
suggested in Kolmogorov as one of possible randomness definitions.) We also get
another 2-randomness criterion by Miller and Nies: is 2-random if and only
if for some and infinitely many prefixes of .
This is a modified version of our old paper that contained a weaker (and
cumbersome) version of Conidis' result, and the proof used low basis theorem
(in quite a strange way). The full version was formulated there as a
conjecture. This conjecture was later proved by Conidis. Bruno Bauwens
(personal communication) noted that the proof can be obtained also by a simple
modification of our original argument, and we reproduce Bauwens' argument with
his permission.Comment: See http://arxiv.org/abs/0802.2833 for the old pape
Test Martingales, Bayes Factors and -Values
A nonnegative martingale with initial value equal to one measures evidence
against a probabilistic hypothesis. The inverse of its value at some stopping
time can be interpreted as a Bayes factor. If we exaggerate the evidence by
considering the largest value attained so far by such a martingale, the
exaggeration will be limited, and there are systematic ways to eliminate it.
The inverse of the exaggerated value at some stopping time can be interpreted
as a -value. We give a simple characterization of all increasing functions
that eliminate the exaggeration.Comment: Published in at http://dx.doi.org/10.1214/10-STS347 the Statistical
Science (http://www.imstat.org/sts/) by the Institute of Mathematical
Statistics (http://www.imstat.org
Test martingales, Bayes factors, and p-values
Project web site
Independent minimum length programs to translate between given strings
AbstractA string p is called a program to compute y given x if U(p,x)=y, where U denotes universal programming language. Kolmogorov complexity K(y|x) of y relative to x is defined as minimum length of a program to compute y given x. Let K(x) denote K(x|emptystring) (Kolmogorov complexity of x) and let I(x:y)=K(x)+K(y)−K(〈x,y〉) (the amount of mutual information in x,y). In the present paper, we answer in the negative the following question posed in Bennett et al., IEEE Trans. Inform. Theory 44 (4) (1998) 1407–1423. Is it true that for any strings x,y there are independent minimum length programs p,q to translate between x,y, that is, is it true that for any x,y there are p,q such that U(p,x)=y, U(q,y)=x, the length of p is K(y|x), the length of q is K(x|y), and I(p:q)=0 (where the last three equalities hold up to an additive O(log(K(x|y)+K(y|x))) term)?
Probability-free pricing of adjusted American lookbacks
Consider an American option that pays G(X^*_t) when exercised at time t,
where G is a positive increasing function, X^*_t := \sup_{s\le t}X_s, and X_s
is the price of the underlying security at time s. Assuming zero interest
rates, we show that the seller of this option can hedge his position by trading
in the underlying security if he begins with initial capital
X_0\int_{X_0}^{\infty}G(x)x^{-2}dx (and this is the smallest initial capital
that allows him to hedge his position). This leads to strategies for trading
that are always competitive both with a given strategy's current performance
and, to a somewhat lesser degree, with its best performance so far. It also
leads to methods of statistical testing that avoid sacrificing too much of the
maximum statistical significance that they achieve in the course of
accumulating data.Comment: 28 pages, 1 figur
Randomized Boolean Decision Trees: Several Remarks
Assume we want to show that (a) the cost of any randomized decision tree computing a given Boolean function is at least c. To this end it suffices to prove that (b) there is a probability distribution over the set of all assignments to variables of that function with respect to which the average cost of any deterministic decision tree computing that function is at least c. Yao in [11] showed that this method is universal for proving lower bounds for randomized errorless decision trees, that is, that (a) is equivalent to (b). In the present paper we prove that this is the case also for randomized decision trees which are allowed to make errors. This gives the positive answer to the question posed in [11]. In the second part of the paper we exhibit an example when randomized directional decision trees (defined in [7]) to evaluate read once formulae are not optimal. We construct a formula Fn of n Boolean variables such that the cost of the optimal directional decision tree computing Fn is..
On Algorithmic Rate-Distortion Function
Abstract — We develop rate-distortion theory in the Kolmogorov complexity setting. This is a theory of lossy compression of individual data objects, using the computable regularities of the data. I
Relationships between NP-sets, Co-NP-sets, and P-sets relative to random oracles
In the present paper we prove that relative to random oracle A (with respect to the uniform measure) the following three assertions hold: (1) there is a pair of disjoint NP A -sets which are separable by no P A - set, (2) there is a pair of disjoint Co-NP A -sets which are separable by no P A -set and (3) there is an infinite Co-NP A -set having no infinite NP A -subset 1 Introduction Many important problems in Complexity theory remain open. The most known one is whether the classes P and NP are equal. It is also unknown if the class NP coincides with the class Co-NP and if NP " Co-NP = P: In the paper [1] it was shown that all these problems have no relativizable solutions. More exactly, oracles A and B were constructed such that P A = NP A (and, therefore, P A = NP A = NP A " Co-NP A ) and NP B 6= Co-NP B (and, therefore, P B 6= Co-NP B ). Using the same technique one can construct an oracle C for which NP C " Co-NP C 6= P C . As the rela..
- …